"main.py --dataset amazon-ratings --hidden_dim 8192 --epoch 18 --lr 0.000098 --runtime 1 --classifier_time 10 --classifier_epochs 5000 --cuda 3 --classifier_lr 0.001 --classifier_decay 0.000005 --batch_norm --loss cosmean --a1 0.2 --a2 0.8 --classifier_seed 0 --GCN_layers_1 1"
test_acc: 0.4542054548423976
test_std: 0.00466442709675041
test_acc_mean: 0.4542054548423976
test_acc_std: 0.0
{
  "classifier_epochs": 5000,
  "batch_size": 1,
  "patience": 200,
  "l2_coef": 0.0,
  "drop_prob": 0.0,
  "dataset": "amazon-ratings",
  "num_hop": 0,
  "cuda": 3,
  "seed": 1234,
  "classifier_lr": 0.001,
  "classifier_decay": 5e-06,
  "classifier_seed": 0,
  "classifier_time": 10,
  "classifier": null,
  "hidden_dim": 8192,
  "proj_dim": 256,
  "fc_out": false,
  "input_dropout": 0.4,
  "dropout": 0.4,
  "batch_norm": true,
  "optimizer": "adam",
  "lr": 9.8e-05,
  "decay": 0.0005,
  "epoch": 18,
  "alpha_dim": "sc",
  "no_alpha_sigmoid": false,
  "beta_dim": "sc",
  "block": "constant",
  "function": "laplacian",
  "use_mlp": false,
  "add_source": false,
  "time": 3.0,
  "augment": false,
  "method": "euler",
  "step_size": 1.0,
  "max_iters": 100,
  "adjoint_method": "adaptive_heun",
  "adjoint": false,
  "adjoint_step_size": 1,
  "tol_scale": 1.0,
  "tol_scale_adjoint": 1.0,
  "ode_blocks": 1,
  "max_nfe": 1000,
  "no_early": false,
  "earlystopxT": 3,
  "max_test_steps": 100,
  "beltrami": false,
  "leaky_relu_slope": 0.2,
  "attention_dropout": 0.0,
  "heads": 4,
  "attention_norm_idx": 0,
  "attention_dim": 16,
  "mix_features": false,
  "reweight_attention": false,
  "attention_type": "scaled_dot",
  "square_plus": false,
  "data_norm": "gcn",
  "self_loop_weight": 1.0,
  "runtime": 1,
  "loss": "cosmean",
  "random_splits": false,
  "alpha_ode": 1.0,
  "alpha_ode_1": 1.0,
  "alpha_ode_2": 0.1,
  "a1": 0.2,
  "a2": 0.8,
  "reaction_term": "bspm",
  "beta_diag": false,
  "lambda": 0.1,
  "GCN_layers_1": 1,
  "GCN_layers_2": 2,
  "num_nodes": 24492
}
